Modelling Motivation as an Intrinsic Reward Signal for Reinforcement Learning Agents
نویسنده
چکیده
Reinforcement learning agents require a learning stimulus in the form of a reward signal in order for learning to occur. Typically, this reward signal makes specific assumptions about the agent’s external environment, such as the presence of certain tasks which should be learned or the presence of a teacher to provide reward feedback. For many complex, dynamic environments, design time knowledge of the tasks to be learned, or the presence of a teacher, cannot be assumed. In order to extend reinforcement learning to such environments, this paper presents a model of motivation as an intrinsic reward signal based on the concept of events, which relaxes these assumptions. The model uses context-free grammars as an adaptable representation of environments about which there is limited design time knowledge, and events to represent potential learning tasks as changes in the agent’s environment. Within this framework, we evaluate a computational model of interest as a motivation process. This evaluation is performed in two reinforcement learning settings, flat reinforcement learning and hierarchical reinforcement learning, in terms of learning efficiency, behavioural variety and behavioural complexity. We show that motivation based on general, task-independent concepts are able to motivate learning of multiple, task-oriented behaviours in environments where neither design time knowledge of the tasks to be learned nor a teacher is available.
منابع مشابه
Exploring Effects of Intrinsic Motivation in Reinforcement Learning Agents
We explore Intrinsic Motivation as a reward framework for learning how to perform complicated tasks. Most reinforcement learning tasks assume the existence of a critic who rewards the agent for its actions. However, taking inspiration for biological agents, we can say that the real critic is the agent itself. We experiment with a model where the rewards are generated by the agent using a proces...
متن کاملWhich is the best intrinsic motivation signal for learning multiple skills?
Humans and other biological agents are able to autonomously learn and cache different skills in the absence of any biological pressure or any assigned task. In this respect, Intrinsic Motivations (i.e., motivations not connected to reward-related stimuli) play a cardinal role in animal learning, and can be considered as a fundamental tool for developing more autonomous and more adaptive artific...
متن کاملHierarchical Functional Concepts for Knowledge Transfer among Reinforcement Learning Agents
This article introduces the notions of functional space and concept as a way of knowledge representation and abstraction for Reinforcement Learning agents. These definitions are used as a tool of knowledge transfer among agents. The agents are assumed to be heterogeneous; they have different state spaces but share a same dynamic, reward and action space. In other words, the agents are assumed t...
متن کاملIntrinsic Motivation and Reinforcement Learning
Psychologists distinguish between extrinsically motivated behavior, which is behavior undertaken to achieve some externally supplied reward, such as a prize, a high grade, or a high-paying job, and intrinsically motivated behavior, which is behavior done for its own sake. Is an analogous distinction meaningful for machine learning systems? Can we say of a machine learning system that it is moti...
متن کاملIncremental learning of skill collections based on intrinsic motivation
Life-long learning of reusable, versatile skills is a key prerequisite for embodied agents that act in a complex, dynamic environment and are faced with different tasks over their lifetime. We address the question of how an agent can learn useful skills efficiently during a developmental period, i.e., when no task is imposed on him and no external reward signal is provided. Learning of skills i...
متن کامل